Goto

Collaborating Authors

 adaptation step


Learning Large-scale Neural Fields via Context Pruned Meta-Learning

Neural Information Processing Systems

We introduce an efficient optimization-based meta-learning technique for large-scale neural field training by realizing significant memory savings through automated online context point selection.


MGDD: A Meta Generator for Fast Dataset Distillation

Neural Information Processing Systems

The meta generator is termed as MGDD in our approach. Once adapted, it can handle arbitrary sizes of synthetic datasets, even for those unseen during adaptation.




HSVA: Hierarchical Semantic-Visual Adaptation for Zero-Shot Learning

Neural Information Processing Systems

Zero-shot learning (ZSL) tackles the unseen class recognition problem, transferring semantic knowledge from seen classes to unseen ones. Typically, to guarantee desirable knowledge transfer, a common (latent) space is adopted for associating the visual and semantic domains in ZSL. However, existing common space learning methods align the semantic and visual domains by merely mitigating distribution disagreement through one-step adaptation. This strategy is usually ineffective due to the heterogeneous nature of the feature representations in the two domains, which intrinsically contain both distribution and structure variations. To address this and advance ZSL, we propose a novel hierarchical semantic-visual adaptation (HSVA) framework.


On Enforcing Better Conditioned Meta-Learning for Rapid Few-Shot Adaptation

Neural Information Processing Systems

Inspired by the concept of preconditioning, we propose a novel method to increase adaptation speed for gradient-based meta-learning methods without incurring extra parameters. We demonstrate that recasting the optimisation problem to a non-linear least-squares formulation provides a principled way to actively enforce a well-conditioned parameter space for meta-learning models based on the concepts of the condition number and local curvature. Our comprehensive evaluations show that the proposed method significantly outperforms its unconstrained counterpart especially during initial adaptation steps, while achieving comparable or better overall results on several few-shot classification tasks - creating the possibility of dynamically choosing the number of adaptation steps at inference time.


Navigating High Dimensional Concept Space with Metalearning

Gupta, Max

arXiv.org Artificial Intelligence

Rapidly learning abstract concepts from limited examples is a hallmark of human intelligence. This work investigates whether gradient-based meta-learning can equip neural networks with inductive biases for efficient few-shot acquisition of discrete concepts. I compare meta-learning methods against a supervised learning baseline on Boolean concepts (logical statements) generated by a probabilistic context-free grammar (PCFG). By systematically varying concept dimensionality (number of features) and recursive compositionality (depth of grammar recursion), I delineate between complexity regimes in which meta-learning robustly improves few-shot concept learning and regimes in which it does not. Meta-learners are much better able to handle compositional complexity than featural complexity. I highlight some reasons for this with a representational analysis of the weights of meta-learners and a loss landscape analysis demonstrating how featural complexity increases the roughness of loss trajectories, allowing curvature-aware optimization to be more effective than first-order methods. I find improvements in out-of-distribution generalization on complex concepts by increasing the number of adaptation steps in meta-SGD, where adaptation acts as a way of encouraging exploration of rougher loss basins. Overall, this work highlights the intricacies of learning compositional versus featural complexity in high dimensional concept spaces and provides a road to understanding the role of 2nd order methods and extended gradient adaptation in few-shot concept learning.


Learning Large-scale Neural Fields via Context Pruned Meta-Learning

Neural Information Processing Systems

We introduce an efficient optimization-based meta-learning technique for large-scale neural field training by realizing significant memory savings through automated online context point selection.



Appendix

Neural Information Processing Systems

This appendix contains the supplementary material for the main text. In Appendix B, we provide details of the derivation of the implicit gradients in Eqs. We first establish here this result. This negative result illustrates the need for alternative estimators. In Appendix A.2.3, we analyze the behavior of Lemma 1 Alternatively, we can follow the approach described in Section 4.2 to estimate both From Eq. (10), we follow the same derivation as Eq.